Goto

Collaborating Authors

 chat translation


An Analysis on Automated Metrics for Evaluating Japanese-English Chat Translation

Rusli, Andre, Shishido, Makoto

arXiv.org Artificial Intelligence

This paper analyses how traditional baseline metrics, such as BLEU and TER, and neural-based methods, such as BERTScore and COMET, score several NMT models performance on chat translation and how these metrics perform when compared to human-annotated scores. The results show that for ranking NMT models in chat translations, all metrics seem consistent in deciding which model outperforms the others. This implies that traditional baseline metrics, which are faster and simpler to use, can still be helpful. On the other hand, when it comes to better correlation with human judgment, neural-based metrics outperform traditional metrics, with COMET achieving the highest correlation with the human-annotated score on a chat translation. However, we show that even the best metric struggles when scoring English translations from sentences with anaphoric zero-pronoun in Japanese.


Exploring the traditional NMT model and Large Language Model for chat translation

Yang, Jinlong, Shang, Hengchao, Wei, Daimeng, Guo, Jiaxin, Li, Zongyao, Wu, Zhanglin, Rao, Zhiqiang, Li, Shaojun, Xie, Yuhao, Luo, Yuanchang, Zheng, Jiawei, Wei, Bin, Yang, Hao

arXiv.org Artificial Intelligence

This paper describes the submissions of Huawei Translation Services Center(HW-TSC) to WMT24 chat translation shared task on English$\leftrightarrow$Germany (en-de) bidirection. The experiments involved fine-tuning models using chat data and exploring various strategies, including Minimum Bayesian Risk (MBR) decoding and self-training. The results show significant performance improvements in certain directions, with the MBR self-training method achieving the best results. The Large Language Model also discusses the challenges and potential avenues for further research in the field of chat translation.


MQM-Chat: Multidimensional Quality Metrics for Chat Translation

Li, Yunmeng, Suzuki, Jun, Morishita, Makoto, Abe, Kaori, Inui, Kentaro

arXiv.org Artificial Intelligence

The complexities of chats pose significant challenges for machine translation models. Recognizing the need for a precise evaluation metric to address the issues of chat translation, this study introduces Multidimensional Quality Metrics for Chat Translation (MQM-Chat). Through the experiments of five models using MQM-Chat, we observed that all models generated certain fundamental errors, while each of them has different shortcomings, such as omission, overly correcting ambiguous source content, and buzzword issues, resulting in the loss of stylized information. Our findings underscore the effectiveness of MQM-Chat in evaluating chat translation, emphasizing the importance of stylized content and dialogue consistency for future studies.


An Investigation of Warning Erroneous Chat Translations in Cross-lingual Communication

Li, Yunmeng, Suzuki, Jun, Morishita, Makoto, Abe, Kaori, Inui, Kentaro

arXiv.org Artificial Intelligence

The complexities of chats pose significant challenges for machine translation models. Recognizing the need for a precise evaluation metric to address the issues of chat translation, this study introduces Multidimensional Quality Metrics for Chat Translation (MQM-Chat). Through the experiments of five models using MQM-Chat, we observed that all models generated certain fundamental errors, while each of them has different shortcomings, such as omission, overly correcting ambiguous source content, and buzzword issues, resulting in the loss of stylized information. Our findings underscore the effectiveness of MQM-Chat in evaluating chat translation, emphasizing the importance of stylized content and dialogue consistency for future studies.


Unified Model Learning for Various Neural Machine Translation

Liang, Yunlong, Meng, Fandong, Xu, Jinan, Wang, Jiaan, Chen, Yufeng, Zhou, Jie

arXiv.org Artificial Intelligence

Existing neural machine translation (NMT) studies mainly focus on developing dataset-specific models based on data from different tasks (e.g., document translation and chat translation). Although the dataset-specific models have achieved impressive performance, it is cumbersome as each dataset demands a model to be designed, trained, and stored. In this work, we aim to unify these translation tasks into a more general setting. Specifically, we propose a ``versatile'' model, i.e., the Unified Model Learning for NMT (UMLNMT) that works with data from different tasks, and can translate well in multiple settings simultaneously, and theoretically it can be as many as possible. Through unified learning, UMLNMT is able to jointly train across multiple tasks, implementing intelligent on-demand translation. On seven widely-used translation tasks, including sentence translation, document translation, and chat translation, our UMLNMT results in substantial improvements over dataset-specific models with significantly reduced model deployment costs. Furthermore, UMLNMT can achieve competitive or better performance than state-of-the-art dataset-specific methods. Human evaluation and in-depth analysis also demonstrate the superiority of our approach on generating diverse and high-quality translations. Additionally, we provide a new genre translation dataset about famous aphorisms with 186k Chinese->English sentence pairs.


BJTU-WeChat's Systems for the WMT22 Chat Translation Task

Liang, Yunlong, Meng, Fandong, Xu, Jinan, Chen, Yufeng, Zhou, Jie

arXiv.org Artificial Intelligence

This paper introduces the joint submission of the Beijing Jiaotong University and WeChat AI to the WMT'22 chat translation task for English-German. Based on the Transformer, we apply several effective variants. In our experiments, we utilize the pre-training-then-fine-tuning paradigm. In the first pre-training stage, we employ data filtering and synthetic data generation (i.e., back-translation, forward-translation, and knowledge distillation). In the second fine-tuning stage, we investigate speaker-aware in-domain data generation, speaker adaptation, prompt-based context modeling, target denoising fine-tuning, and boosted self-COMET-based model ensemble. Our systems achieve 0.810 and 0.946 COMET scores. The COMET scores of English-German and German-English are the highest among all submissions.


Chat Translation: NTT Docomo Debuts First Speech-to-Speech Translation App

AITopics Original Links

If you've ever been struggling with a foreign language dictionary abroad, wishing that you could simply speak into a machine and have your chat translated for you, NTT (News - Alert) Docomo may be ready to make your wish come true. The company, which is Japan's number one cell phone carrier is about to begin offering a new real-time speech-to-speech translation service that you can use both in person and over the phone during a call, according to Geek.com According to Japanese news services, the solution is the first automated chat translation service in the world that is available on a standard cell phone. The new product combines several cutting-edge technologies: advanced speech recognition, machine translation and text-to-speech conversion of the translated results, says Geek.com. The services to power the solution will be cloud-based, says NTT Docomo.